Voice-driven animation

نویسندگان

  • Matthew Brand
  • Ken Shan
چکیده

We introduce a method for learning a mapping between signals, and use this to drive facial animation directly from vocal cues. Instead of depending on heuristic intermediate representations such as phonemes or visemes, the system learns its own representation, which includes dynamical and contextual information. In principle, this allows the system to make optimal use of context to handle ambiguity and relatively long-lasting facial co-articulation effects. The output is a series of facial control parameters, suitable for driving many different kinds of animation ranging from photo-realistic image warps to 3D cartoon characters. This work may not be copied or reproduced in whole or in part for any commercial purpose. Permission to copy in whole or in part without payment of fee is granted for nonprofit educational and research purposes provided that all such whole or partial copies include the following: a notice that such copying is by permission of Mitsubishi Electric Information Technology Center America; an acknowledgment of the authors and individual contributions to the work; and all applicable portions of the copyright notice. Copying, reproduction, or republishing for any other purpose shall require a license with payment of fee to Mitsubishi Electric Information Technology Center America. All rights reserved. Copyright c Mitsubishi Electric Information Technology Center America, 1998 201 Broadway, Cambridge, Massachusetts 02139 Publication History:– 1. 6jul98 first circulated. 2. 26aug98 accepted to Workshop on Perceptual User Interfaces, November 1998, San Francisco 3. 21sep98 final version uploaded to workshop web site. Voice-driven animation Matthew Brand and Ken Shan MERL—a Mitsubishi Electric Research Lab 201 Broadway, Cambridge, MA, 02139 [email protected]

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Design and Implementation of a Voice-Driven Animation System

This thesis presents a novel multimodal interface for directing the actions of computer animated characters and camera movements. Our system can recognize human voice input combined with mouse pointing to generate desired character animation based on motion capture data. We compare our voice-driven system with a button-driven animation interface that has equivalent capabilities. An informal use...

متن کامل

Laughing, Crying, Sneezing and Yawning: Automatic Voice Driven Animation of Non-Speech Articulations

In this paper a technique is presented for learning audio-visual correlations in non-speech related articulations such as laughs, cries, sneezes and yawns, such that accurate new visual motions may be created given just audio. Our underlying model is data-driven and provides reliable performance given voices the system is familiar with as well as new voices. We demonstrate how performance accur...

متن کامل

Real-time Voice Driven Facial Animation System - Systems, Man, and Cybernetics, 1999. IEEE SMC '99 Conference Proceedings. 1999 IEEE International

Recently computer can make cyberspace to walk through by an interactive virtual reality technique. An avatar in cyberspace can bring us a virtual face-to-face communication environment. In this paper, we realize an avatar which has a real face in cyberspace to construct a multi-user -communication system by voice transmission through network. Voice from microphone is transmitted and analyzed, t...

متن کامل

Real-time Face Animation Driven by Human Voice

-This paper presents a review of the recent research that examines the problem of generating full facial animation from auditory input speech signal. The major applications of this technique include face animation, human-computer interfaces, computer-aided instruction, video games and multimedia telephony for the hearing impaired. Using human voice for the real-time animation of synthetic faces...

متن کامل

Text2Video: Text-Driven Facial Animation using MPEG-4

We present a complete system for the automatic creation of talking head video sequences from text messages. Our system converts the text into MPEG-4 Facial Animation Parameters and synthetic voice. A user selected 3D character will perform lip movements synchronized to the speech data. The 3D models created from a single image vary from realistic people to cartoon characters. A voice selection ...

متن کامل

Dynamic Mapping Method Based Speech Driven Face Animation System

In the paper, we design and develop a speech driven face animation system based on the dynamic mapping method. The face animation is synthesized by the unit concatenating, and synchronous with the real speech. The units are selected according to the cost functions which correspond to voice spectrum distance between training and target units. Visual distance between two adjacent training units i...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 1998